Sort by
Parameterization of manipulative media discourse: possibilities and problems of automatic diagnosis

Abstract The issue of quantitative measurement and automatic processing is a significant problem in determining the markers of the manipulative potential of media texts, since linguistic indicators are the basis of machine parameterization. The purpose of the research is to analyse the possibilities of the main language parameters of the manipulativeness of media discourse, which can be identified using machine learning. To achieve the research goals, the following methods were used: system, content analysis, computer modelling, and comparative. The results of the article determined that such language indicators as use of the subjunctive mood of verbs, capital letters, high frequency of use of the ‘not’ particle, punctuation marks, questions, or exclamations of a rhetorical nature, use of quotation marks for the purpose of irony, double negative sentences, use of the word ‘no’, and verbal structures calling to action act as computer classification parameters. In order to cover the above purpose, PYTHON software was implemented that allowed texts to be analysed and visualized in algorithmic and lexical-vocabulary ways. In addition, it was determined that by integrating the PYTHON tool, it became possible to use language transformation markers that formed linguistic patterns in the analysed text. The list of parameters for diagnosing manipulative texts is non-exhaustive, which emphasizes the possibility of machine measurement of the manipulative component of mass media discourse.

Just Published
Relevant
A statistical approach to Hollywood remake and sequel metadata

Abstract Hollywood film remakes, as old as the cinema itself, have attracted much professional, critical, and academic attention. They have been viewed by art critics as products of cultural derivativity and imperialism and commended by financial experts as low-risk business investments, closely linked to other forms of brand extension, such as sequels and bestseller adaptations. In this article, we adopt a film-historical quantitative approach to Hollywood film remakes by analysing metadata obtained from the Internet Movie Database (IMDb) and verified against reliable print and web sources. We analyse 986 Hollywood remakes produced between 1915 and 2020 in terms of raw and relative frequencies of annual releases, genre (in)stability, and patterns of transnational reproduction. We contrast our findings with those outlined by Henderson (2014a) in his statistical survey of Hollywood sequels, series films, prequels, and spin-offs, presented in his monograph The Hollywood Sequel: History and Form, 1911–2010. Having completed his list with recent sequential productions released between 2011 and 2020, we investigate the potential parallels between Hollywood remaking and sequelization practices. Our findings demonstrate historical discrepancies in various ‘content recycling’ trends, which help better characterize the cultural and commercial significance of remakes and serial forms in the American film industry.

Just Published
Relevant
Is medieval distant viewing possible? : Extending and enriching annotation of legacy image collections using visual analytics

Abstract Distant viewing approaches have typically used image datasets close to the contemporary image data used to train machine learning models. To work with images from other historical periods requires expert annotated data, and the quality of labels is crucial for the quality of results. Especially when working with cultural heritage collections that contain myriad uncertainties, annotating data, or re-annotating, legacy data is an arduous task. In this paper, we describe working with two pre-annotated sets of medieval manuscript images that exhibit conflicting and overlapping metadata. Since a manual reconciliation of the two legacy ontologies would be very expensive, we aim (1) to create a more uniform set of descriptive labels to serve as a “bridge” in the combined dataset, and (2) to establish a high-quality hierarchical classification that can be used as a valuable input for subsequent supervised machine learning. To achieve these goals, we developed visualization and interaction mechanisms, enabling medievalists to combine, regularize and extend the vocabulary used to describe these, and other cognate, image datasets. The visual interfaces provide experts an overview of relationships in the data going beyond the sum total of the metadata. Word and image embeddings as well as co-occurrences of labels across the datasets enable batch re-annotation of images, recommendation of label candidates, and support composing a hierarchical classification of labels.

Just Published
Relevant
Are Digital Humanities platforms facilitating sufficient diversity in research? A study of the Transkribus Scholarship Programme

Abstract To what extent do Digital Humanities (DH) platforms support access to diverse user cohorts? We take, as a case study, the Automated Text Recognition (ATR) platform Transkribus and its Transkribus Scholarship Programme (TSP), which provides free processing credits to eligible users. Using a mixed methods approach we address the following questions: What are the demographics of those using the TSP scheme? What work is enabled by such a scheme? How can this inform more equitable access to DH platforms? The findings demonstrate how ATR tools are currently used and made accessible. TSP applicants are overwhelmingly students (n = 111/156, 71.15 per cent) drawn from universities and research institutes, mostly in Europe, but are globally distributed; representing institutions that do not hold shares in Transkribus, and indicating a diverse user pipeline. Further work is required to increase potential benefits of the scholarship and to ensure sustainability. Increased dialogue between the Recognition and Enrichment of Archival Document-COOP and applicants would assist in the calculation of processing costs. We show financial—or in-kind—support is necessary to increase access to paid-for platforms, ensuring a diversity of DH research. We also provide recommendations for platform providers and funding bodies regarding access and the impact this can have, including locating a sustainable balance between absorbing the costs of maintaining DH or digital scholarship tools and providing sufficient support and training to further enable diverse research.

Open Access
Relevant
Tang Chang’an poetry automatic classification: a practical application of deep learning methods

Abstract As the capital of Tang Dynasty, Chang’an was one of the most prosperous cities in the world at that time and had a profound influence on Tang poetry. Poets described Chang’an to illustrate the cultural features of the Tang Dynasty while also invoking emotions in readers. The study of Tang Chang’an poetry has important literary and historical value. In order to understand the interpretation and emotional expression of Tang Chang’an poetry more conveniently and clearly, we conducted a study using deep learning to classify Chang’an poetry into four classes: imperially assigned poetry (应制), emotional poetry (感怀), parting poetry (离别), and other poetry (其他). We suggested a comprehensive framework of text classification based on deep learning, including a text input module, feature encoder module, and classification module. We applied several mainstream deep neural network structures to extract features in different ways, which comprised convolutional neural network (CNN), Fasttext, bi-direction long-short-term memory network, and Attention mechanism. Based on our experimental findings, the CNN-based method achieved the best performance for the task. Our inference was that, in Chinese ancient poetry, the analysis of semantic content is more facilitated by local textual features rather than contextual features. We combined this inference with the theory of image in Chinese ancient poetry to analyze the suitability of the deep learning techniques for the study of Chinese ancient poetry.

Open Access
Relevant
Modelling Chinese contemporary calligraphy: the WRITE data model

Abstract This article presents the WRITE data model and dataset, a comprehensive collection of Chinese contemporary calligraphic data, utilizing Linked Open Data (LOD) principles. Calligraphy plays a pivotal role in Chinese culture, reflecting national identity and cultural transformations. The objective of this study is to enhance understanding and provide new tools for exploring Chinese contemporary calligraphy through LOD. The WRITE data model comprises artistic, linguistic, and socio-political-economic aspects. The WRITE data model, developed collaboratively with domain specialists, represents four collections: Contemporary Visual Art, Performance, Graffiti, and Decorative and Applied Arts. Metadata describing the artworks is structured by reusing and extending the Wikidata model. Complex relations are established between artworks and contextual elements, (e.g. people, exhibition history, organizations, and literary works). The artistic and linguistic metadata recorded over the ‘calli-writing units’ provide insights into shared and diverging characteristics with traditional calligraphy. Traditional and contemporary calligraphy practices are compared, highlighting how contemporary calligraphy challenges traditional rules. Two case studies demonstrate the formalization of specific items in the WRITE collection, showcasing the study of graffiti art’s socio-political meaning in China and the multidimensional nature of musicalligraphy performance. The WRITE dataset and data model contribute to advancing knowledge and understanding of Chinese contemporary calligraphy, offering valuable resources for artistic analysis and interdisciplinary research.

Open Access
Relevant